首页> 外文OA文献 >Statistical consistency and asymptotic normality for high-dimensional robust M-estimators
【2h】

Statistical consistency and asymptotic normality for high-dimensional robust M-estimators

机译:高维数据的统计一致性和渐近正态性   强大的m估计

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

We study theoretical properties of regularized robust M-estimators,applicable when data are drawn from a sparse high-dimensional linear model andcontaminated by heavy-tailed distributions and/or outliers in the additiveerrors and covariates. We first establish a form of local statisticalconsistency for the penalized regression estimators under fairly mildconditions on the error distribution: When the derivative of the loss functionis bounded and satisfies a local restricted curvature condition, all stationarypoints within a constant radius of the true regression vector converge at theminimax rate enjoyed by the Lasso with sub-Gaussian errors. When an appropriatenonconvex regularizer is used in place of an l_1-penalty, we show that suchstationary points are in fact unique and equal to the local oracle solutionwith the correct support---hence, results on asymptotic normality in thelow-dimensional case carry over immediately to the high-dimensional setting.This has important implications for the efficiency of regularized nonconvexM-estimators when the errors are heavy-tailed. Our analysis of the localcurvature of the loss function also has useful consequences for optimizationwhen the robust regression function and/or regularizer is nonconvex and theobjective function possesses stationary points outside the local region. Weshow that as long as a composite gradient descent algorithm is initializedwithin a constant radius of the true regression vector, successive iterateswill converge at a linear rate to a stationary point within the local region.Furthermore, the global optimum of a convex regularized robust regressionfunction may be used to obtain a suitable initialization. The result is a noveltwo-step procedure that uses a convex M-estimator to achieve consistency and anonconvex M-estimator to increase efficiency.
机译:我们研究了正则化鲁棒M估计量的理论性质,适用于从稀疏高维线性模型中提取数据并被加法误差和协变量中的重尾分布和/或离群值污染的情况。我们首先在误差分布的相当温和的条件下为惩罚回归估计量建立一种局部统计一致性形式:当损失函数的导数有界并满足局部受限曲率条件时,真实回归向量的恒定半径内的所有静止点都收敛于套索具有次高斯误差的最小极小率。当使用适当的非凸正则化函数代替l_1惩罚时,我们证明了这些平稳点实际上是唯一的,并且与带有正确支持的本地oracle解决方案相等-因此,在低维情况下,渐近正态性的结果立即结转误差较大时,这对正则化非凸M估计量的效率具有重要意义。当鲁棒回归函数和/或正则化函数是非凸的并且目标函数在局部区域之外具有固定点时,我们对损失函数的局部曲率的分析也对优化产生有用的结果。我们证明,只要在真实回归向量的恒定半径内初始化复合梯度下降算法,连续迭代将以线性速率收敛到局部区域内的固定点。此外,凸正则化鲁棒回归函数的全局最优值可能是用于获得合适的初始化。结果是一个新颖的两步过程,该过程使用凸M估计量来实现一致性,并使用非凸M估计量来提高效率。

著录项

  • 作者

    Loh, Po-Ling;

  • 作者单位
  • 年度 2015
  • 总页数
  • 原文格式 PDF
  • 正文语种
  • 中图分类

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号